Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@josemurilo@mato.social
2024-02-29 16:42:10

“A sign of the scale of #’Twitter’s bot problem is the thriving #botindustry. Bot makers from around the world advertise their services on freelancer websites.
A computer scientist in Pakistan, sells "ChatGPT Twitter bots" for $30 to $500, depending on complexity.
In an interview with the ABC …he said could make a "fully fledged" bot that could "like comments o…

@chris@mstdn.chrisalemany.ca
2024-02-29 16:53:55

*sarc
Sounds great over there!
P.S. Twitter is NOT the “Internet” fgs….
#TwitterEnding #Twitter #Mastodon

@crell@phpc.social
2024-02-28 16:33:34

When you set your paperclip optimizers to optimize for "engagement," this is the obvious, inevitable result.
abc.net.au/news/science/2024-0

@arXiv_csCL_bot@mastoxiv.page
2024-05-01 06:49:12

Better & Faster Large Language Models via Multi-token Prediction
Fabian Gloeckle, Badr Youbi Idrissi, Baptiste Rozi\`ere, David Lopez-Paz, Gabriel Synnaeve
arxiv.org/abs/2404.19737 arxiv.org/pdf/2404.19737
arXiv:2404.19737v1 Announce Type: new
Abstract: Large language models such as GPT and Llama are trained with a next-token prediction loss. In this work, we suggest that training language models to predict multiple future tokens at once results in higher sample efficiency. More specifically, at each position in the training corpus, we ask the model to predict the following n tokens using n independent output heads, operating on top of a shared model trunk. Considering multi-token prediction as an auxiliary training task, we measure improved downstream capabilities with no overhead in training time for both code and natural language models. The method is increasingly useful for larger model sizes, and keeps its appeal when training for multiple epochs. Gains are especially pronounced on generative benchmarks like coding, where our models consistently outperform strong baselines by several percentage points. Our 13B parameter models solves 12 % more problems on HumanEval and 17 % more on MBPP than comparable next-token models. Experiments on small algorithmic tasks demonstrate that multi-token prediction is favorable for the development of induction heads and algorithmic reasoning capabilities. As an additional benefit, models trained with 4-token prediction are up to 3 times faster at inference, even with large batch sizes.

@arXiv_csDS_bot@mastoxiv.page
2024-04-30 07:20:33

Private graph colouring with limited defectiveness
Aleksander B. G. Christiansen, Eva Rotenberg, Teresa Anna Steiner, Juliette Vlieghe
arxiv.org/abs/2404.18692 arxiv.org/pdf/2404.18692
arXiv:2404.18692v1 Announce Type: new
Abstract: Differential privacy is the gold standard in the problem of privacy preserving data analysis, which is crucial in a wide range of disciplines. Vertex colouring is one of the most fundamental questions about a graph. In this paper, we study the vertex colouring problem in the differentially private setting.
To be edge-differentially private, a colouring algorithm needs to be defective: a colouring is d-defective if a vertex can share a colour with at most d of its neighbours. Without defectiveness, the only differentially private colouring algorithm needs to assign n different colours to the n different vertices. We show the following lower bound for the defectiveness: a differentially private c-edge colouring algorithm of a graph of maximum degree {\Delta} > 0 has defectiveness at least d = {\Omega} (log n / (log c log {\Delta})).
We also present an {\epsilon}-differentially private algorithm to {\Theta} ( {\Delta} / log n 1 / {\epsilon})-colour a graph with defectiveness at most {\Theta}(log n).

@jdrm@social.linux.pizza
2024-02-28 16:31:47

No entiendo a la gente que sigue ahí "porque tiene mucha difusión" abc.net.au/news/science/2024-0

@kernellogger@fosstodon.org
2024-04-27 06:08:10

How Allegro reduced latency outliers by 82% by switching to #XFS:
blog.allegro.tech/2024/03/kafk
"'Using a com…

@arXiv_csNE_bot@mastoxiv.page
2024-05-01 08:36:06

This arxiv.org/abs/2303.00614 has been replaced.
initial toot: mastoxiv.page/@arXiv_csNE_…

@arXiv_csCG_bot@mastoxiv.page
2024-04-30 06:47:35

A faster algorithm for the Fr\'echet distance in 1D for the imbalanced case
Lotte Blank, Anne Driemel
arxiv.org/abs/2404.18738 <…

@zachleat@zachleat.com
2024-03-29 17:00:41

@… @… definitely feeling this! Our ability to grow communities has seen incredible fragmentation in the last year and it’s resulting in stuff being seen a lot less. (Not to mention the rise of algorithmically g…